16 research outputs found
Optimal Planning Modulo Theories
Planning for real-world applications requires algorithms and tools with the ability to handle the complexity such scenarios entail. However, meeting the needs of such applications poses substantial challenges, both representational and algorithmic. On the one hand, expressive languages are needed to build faithful models. On the other hand, efficient solving techniques that can support these languages need to be devised. A response to this challenge is underway, and the past few years witnessed a community effort towards more expressive languages, including decidable fragments of first-order theories. In this work we focus on planning with arithmetic theories and propose Optimal Planning Modulo Theories, a framework that attempts to provide efficient means of dealing with such problems. Leveraging generic Optimization Modulo Theories (OMT) solvers, we first present domain-specific encodings for optimal planning in complex logistic domains. We then present a more general, domain- independent formulation that allows to extend OMT planning to a broader class of well-studied numeric problems in planning. To the best of our knowledge, this is the first time OMT procedures are employed in domain-independent planning
Provably Robust and Plausible Counterfactual Explanations for Neural Networks via Robust Optimisation
Counterfactual Explanations (CEs) have received increasing interest as a
major methodology for explaining neural network classifiers. Usually, CEs for
an input-output pair are defined as data points with minimum distance to the
input that are classified with a different label than the output. To tackle the
established problem that CEs are easily invalidated when model parameters are
updated (e.g. retrained), studies have proposed ways to certify the robustness
of CEs under model parameter changes bounded by a norm ball. However, existing
methods targeting this form of robustness are not sound or complete, and they
may generate implausible CEs, i.e., outliers wrt the training dataset. In fact,
no existing method simultaneously optimises for proximity and plausibility
while preserving robustness guarantees. In this work, we propose Provably
RObust and PLAusible Counterfactual Explanations (PROPLACE), a method
leveraging on robust optimisation techniques to address the aforementioned
limitations in the literature. We formulate an iterative algorithm to compute
provably robust CEs and prove its convergence, soundness and completeness.
Through a comparative experiment involving six baselines, five of which target
robustness, we show that PROPLACE achieves state-of-the-art performances
against metrics on three evaluation aspects.Comment: Accepted at ACML 2023, camera-ready versio
Formalising the Robustness of Counterfactual Explanations for Neural Networks
The use of counterfactual explanations (CFXs) is an increasingly popular explanation strategy for machine learning models. However, recent studies have shown that these explanations may not be robust to changes in the underlying model (e.g., following retraining), which raises questions about their reliability in real-world applications. Existing attempts towards solving this problem are heuristic, and the robustness to model changes of the resulting CFXs is evaluated with only a small number of retrained models, failing to provide exhaustive guarantees. To remedy this, we propose â-robustness, the first notion to formally and deterministically assess the robustness (to model changes) of CFXs for neural networks. We introduce an abstraction framework based on interval neural networks
to verify the â-robustness of CFXs against a possibly infinite set of changes to the model parameters, i.e., weights and biases. We then demonstrate the utility of this approach in two distinct ways. First, we analyse the â-robustness of a number of CFX generation methods from the literature and show that they unanimously host significant deficiencies in this regard. Second, we demonstrate how embedding â-robustness within existing methods can provide CFXs which are provably robust
Engineering Controllers For Swarm Robotics Via Reachability Analysis In Hybrid Systems
In this paper we propose hybrid systems and reachability analysis to verify properties in swarm robotics systems, i.e., teams of robots performing cooperative tasks without any centralized coordination. We discuss the challenges that are to be faced and we report on the experience gained from applying hybrid formalisms to the verification of swarm robotics systems